1,758 research outputs found

    Exploring requirements and detector solutions for FCC-ee

    Full text link
    Circular colliders have the advantage of delivering collisions to multiple interaction points, which allow different detector designs to be studied and optimized - up to four for FCC-ee. On the one hand, the detectors must satisfy the constraints imposed by the invasive interaction region layout. On the other hand, the performance of heavy-flavour tagging, of particle identification, of tracking and particle-flow reconstruction, and of lepton, jet, missing energy and angular resolution, need to match the physics programme and the exquisite statistical precision offered by FCC-ee. During the FCC feasibility study (2021-2025), benchmark physics processes will be used to determine, via appropriate simulations, the requirements on the detector performance or design that must be satisfied to ensure that the systematic uncertainties of the measurements are commensurate with their statistical precision. The usage of the data themselves, in order to reach the challenging goals on the stability and on the alignment of the detector, in particular for the programme at and around the Z peak, will also be studied. In addition, the potential for discovering very weakly coupled new particles, in decays of Z or Higgs bosons, could motivate dedicated detector designs that would increase the efficiency for reconstructing the unusual signatures of such processes. These studies are a crucial input to the further optimization of the two concepts described in the Conceptual Design Report, CLD, and IDEA, and to the development of new concepts which might actually prove to be better adapted to the FCC-ee physics programme, or parts thereof.Comment: Submitted to EPJ+ special issue: A future Higgs and Electroweak factory (FCC): Challenges towards discovery, Focus on FCC-e

    Investigation of in the full hadronic final state at CDF with a neural network approach

    Get PDF
    Abstract In this work we present the results of a neural network (NN) approach to the measurement of the t t production cross-section and top mass in the all-hadronic channel, analyzing data collected at the Collider Detector at Fermilab (CDF) experiment. We have used a hardware implementation of a feed forward neural network, TOTEM , the product of a collaboration of INFN (Istituto Nazionale Fisica Nucleare)—IRST (Istituto per la Ricerca Scientifica e Tecnologica)—University of Trento, Italy. Particular attention has been payed to the evaluation of the systematics specifically related to the NN approach. The results are consistent with those obtained at CDF by conventional data selection techniques

    Prospective Studies for LEP3 with the CMS Detector

    Full text link
    On July 4, 2012, the discovery of a new boson, with mass around 125 GeV/c2 and with properties compatible with those of a standard-model Higgs boson, was announced at CERN. In this context, a high-luminosity electron-positron collider ring, operating in the LHC tunnel at a centre-of-mass energy of 240 GeV and called LEP3, becomes an attractive opportunity both from financial and scientific point of views. The performance and the suitability of the CMS detector are evaluated, with emphasis on an accurate measurement of the Higgs boson properties. The precision expected for the Higgs boson couplings is found to be significantly better than that predicted by Linear Collider studies.Comment: 25 pages, 13 figures, 3 tables, Submitted to the European Strategy Preparatory Group (contribution 171), Companion to arXiV:1208.0504. Version 2 contains several additional Higgs decay channels studied, and numbers for TLEP as well as for LEP

    Tracker Operation and Performance at the Magnet Test and Cosmic Challenge

    Get PDF
    During summer 2006 a fraction of the CMS silicon strip tracker was operated in a comprehensive slice test called the Magnet Test and Cosmic Challenge (MTCC). At the MTCC, cosmic rays detected in the muon chambers were used to trigger the readout of all CMS sub-detectors in the general data acquisition system and in the presence of the 4 T magnetic field produced by the CMS superconducting solenoid. This document describes the operation of the Tracker hardware and software prior, during and after data taking. The performance of the detector as resulting from the MTCC data analysis is also presented

    Report from Working Group 3: Beyond the standard model physics at the HL-LHC and HE-LHC

    Get PDF
    This is the third out of five chapters of the final report [1] of the Workshop on Physics at HL-LHC, and perspectives on HE-LHC [2]. It is devoted to the study of the potential, in the search for Beyond the Standard Model (BSM) physics, of the High Luminosity (HL) phase of the LHC, defined as 33 ab1^{-1} of data taken at a centre-of-mass energy of 14 TeV, and of a possible future upgrade, the High Energy (HE) LHC, defined as 1515 ab1^{-1} of data at a centre-of-mass energy of 27 TeV. We consider a large variety of new physics models, both in a simplified model fashion and in a more model-dependent one. A long list of contributions from the theory and experimental (ATLAS, CMS, LHCb) communities have been collected and merged together to give a complete, wide, and consistent view of future prospects for BSM physics at the considered colliders. On top of the usual standard candles, such as supersymmetric simplified models and resonances, considered for the evaluation of future collider potentials, this report contains results on dark matter and dark sectors, long lived particles, leptoquarks, sterile neutrinos, axion-like particles, heavy scalars, vector-like quarks, and more. Particular attention is placed, especially in the study of the HL-LHC prospects, to the detector upgrades, the assessment of the future systematic uncertainties, and new experimental techniques. The general conclusion is that the HL-LHC, on top of allowing to extend the present LHC mass and coupling reach by 2050%20-50\% on most new physics scenarios, will also be able to constrain, and potentially discover, new physics that is presently unconstrained. Moreover, compared to the HL-LHC, the reach in most observables will, generally more than double at the HE-LHC, which may represent a good candidate future facility for a final test of TeV-scale new physics

    Search for New Particles Decaying to b bbar in p pbar Collisions at sqrt{s}=1.8 TeV

    Full text link
    We have used 87 pb^-1 of data collected with the Collider Detector at Fermilab to search for new particles decaying to b bbar. We present model-independent upper limits on the cross section for narrow resonances which excludes the color-octet technirho in the mass interval 350 < M < 440 GeV/c^2. In addition, we exclude topgluons, predicted in models of topcolor-assisted technicolor, of width Gamma = 0.3 M in the mass range 280 < M < 670 GeV/c^2, of width Gamma = 0.5 M in the mass range 340 < M < 640 GeV/c^2, and of width Gamma = 0.7 M in the mass range 375 < M < 560 GeV/c^2.Comment: 17 pages in a LaTex generated postscript file, with one table and four figures. Resubmitted to Physical Review Letters. Minor clarifications were added to the text. The displayed normalization of the resonance models in Figure 2 was modified to correspond to our 95% CL upper limit on the cross section (instead of arbitrary normalization which was used previously). All results are identical to those in the previous submissio

    Optimasi Portofolio Resiko Menggunakan Model Markowitz MVO Dikaitkan dengan Keterbatasan Manusia dalam Memprediksi Masa Depan dalam Perspektif Al-Qur`an

    Full text link
    Risk portfolio on modern finance has become increasingly technical, requiring the use of sophisticated mathematical tools in both research and practice. Since companies cannot insure themselves completely against risk, as human incompetence in predicting the future precisely that written in Al-Quran surah Luqman verse 34, they have to manage it to yield an optimal portfolio. The objective here is to minimize the variance among all portfolios, or alternatively, to maximize expected return among all portfolios that has at least a certain expected return. Furthermore, this study focuses on optimizing risk portfolio so called Markowitz MVO (Mean-Variance Optimization). Some theoretical frameworks for analysis are arithmetic mean, geometric mean, variance, covariance, linear programming, and quadratic programming. Moreover, finding a minimum variance portfolio produces a convex quadratic programming, that is minimizing the objective function ðð¥with constraintsð ð 𥠥 ðandð´ð¥ = ð. The outcome of this research is the solution of optimal risk portofolio in some investments that could be finished smoothly using MATLAB R2007b software together with its graphic analysis
    corecore